Active Feature Selection for the Mutual Information Criterion
نویسندگان
چکیده
We study active feature selection, a novel selection setting in which unlabeled data is available, but the budget for labels limited, and examples to label can be actively selected by algorithm. focus on using classical mutual information criterion, selects k features with largest label. In setting, goal use significantly fewer than set size still find whose based entire large. explain experimentally choices that we make algorithm, show they lead successful compared other more naive approaches. Our design draws insights relate problem of pure-exploration multi-armed bandits settings. While here information, our general methodology adapted feature-quality measures as well. The extended version this paper, reporting all experiment results, available at Schnapp Sabato (2020). code following url: https://github.com/ShacharSchnapp/ActiveFeatureSelection
منابع مشابه
On the Feature Selection Criterion Proposed in ‘Gait Feature Subset Selection by Mutual Information’
Abstract Recently, Guo and Nixon [1] proposed a feature selection method based on maximizing I(x; Y ), the multidimensional mutual information between feature vector x and class variable Y . Because computing I(x; Y ) can be difficult in practice, Guo and Nixon proposed an approximation of I(x; Y ) as the criterion for feature selection. We show that Guo and Nixon’s criterion originates from ap...
متن کاملQuadratic Mutual Information Feature Selection
We propose a novel feature selection method based on quadratic mutual information which has its roots in Cauchy–Schwarz divergence and Renyi entropy. The method uses the direct estimation of quadratic mutual information from data samples using Gaussian kernel functions, and can detect second order non-linear relations. Its main advantages are: (i) unified analysis of discrete and continuous dat...
متن کاملOn Estimating Mutual Information for Feature Selection
Mutual Information (MI) is a powerful concept from information theory used in many application fields. For practical tasks it is often necessary to estimate the Mutual Information from available data. We compare state of the art methods for estimating MI from continuous data, focusing on the usefulness for the feature selection task. Our results suggest that many methods are practically relevan...
متن کاملWeighted Mutual Information for Feature Selection
In this paper, we apply weighted Mutual Information for effective feature selection. The presented hybrid filter wrapper approach resembles the well known AdaBoost algorithm by focusing on those samples that are not classified or approximated correctly using the selected features. Redundancies and bias of the employed learning machine are handled implicitly by our approach. In experiments, we c...
متن کاملMutual Information Criteria for Feature Selection
In many data analysis tasks, one is often confronted with very high dimensional data. The feature selection problem is essentially a combinatorial optimization problem which is computationally expensive. To overcome this problem it is frequently assumed either that features independently influence the class variable or do so only involving pairwise feature interaction. In prior work [18], we ha...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i11.17144